Goto

Collaborating Authors

 optimization trajectory



On the Limitations of Fractal Dimension as a Measure of Generalization Charlie B. Tan University of Oxford Inés García-Redondo Imperial College London Qiquan Wang

Neural Information Processing Systems

Bounding and predicting the generalization gap of overparameterized neural networks remains a central open problem in theoretical machine learning. There is a recent and growing body of literature that proposes the framework of fractals to model optimization trajectories of neural networks, motivating generalization bounds and measures based on the fractal dimension of the trajectory. Notably, the persistent homology dimension has been proposed to correlate with the generalization gap.



e04101138a3c94544760c1dbdf2c7a2d-Paper-Conference.pdf

Neural Information Processing Systems

For example, while prior work has suggested that theglobally optimal VAEsolution canlearn thecorrect manifold dimension, anecessary (butnotsufficient)condition forproducing samplesfrom the true data distribution, this has never been rigorously proven. Moreover, it remains unclear how such considerations would change when various types of conditioning variablesare introduced, or when the data support is extended to a union of manifolds (e.g., as is likely the case for MNIST digits and related). In this work, we address these points by first proving that VAE global minima are indeed capable of recovering the correct manifold dimension.







addressed all raised questions below: conducting new experiments to compare with hand-designed optimizers (# 1)

Neural Information Processing Systems

We genuinely appreciate all three reviewers' (#1,#2,#3,#4) valuable suggestions to strengthen our paper. More details are referred to Reviewer #3's Q2. We sincerely appreciate your suggestion and will revise the caption in figure 4 for better readability. Reply: It is a great observation. Thus, IL appears to be a main contributor for L2O generalizing across different optimizees.